Learning Shallow Context-free Languages under Simple Distributions
نویسنده
چکیده
In this paper I present the EMILE 3.0 algorithm 1. It can learn shallow context-free grammars eeciently. It does so under circumstances that, from a perspective of complexity, come resonably close to the conditions under which human beings learn a language. A language is shallow in its descriptive length if all the relevant constructions we need to know to learn it have a complexity that is logarithmic in the descriptive length of the grammar as a whole. I claim that natural languages are shallow and that shallowness in itself is an interesting general constraint in the context of formal learning theory. I also introduce the concept of sepa-rability. A language is separable if the validity of rules can be tested by means of memberbership queries using simple substitutions. Although I do not believe that human languages are strictly separable I do think it is an important aspect of eecient human communication. I illustrate the structure of the algorithm by means of an extensive example, and I indicate how a careful analysis of the bias involved in natural language learning may lead to versions of EMILE 3.0 that can be implemented in real-life dialogue systems. I claim that the EMILE approach could serve as a valuable metamodel for evaluating clustering approaches to language learning. The results of EMILE are mainly theoretical. They 1 The original acronym stands for Entity Modeling Intelligent Learning Engine. It refers to earlier versions of EMILE that also had semantic capacities. The name EMILE is also motivated by the book on education by J.-J. Rousseau. In order to distinguishseveral versions of EMILE I started recently to give them version numbers. Versions 1 and 2 are discussed in Adriaans, 1992a. 1 2 / Pieter W. Adriaans could potentially also be used to develop more eecient variants of other approaches to language learning.
منابع مشابه
Partially Distribution-Free Learning of Regular Languages from Positive Samples
Regular languages are widely used in NLP today in spite of their shortcomings. Efficient algorithms that can reliably learn these languages, and which must in realistic applications only use positive samples, are necessary. These languages are not learnable under traditional distribution free criteria. We claim that an appropriate learning framework is PAC learning where the distributions are c...
متن کاملUsing Contextual Representations to Efficiently Learn Context-Free Languages
We present a polynomial update time algorithm for the inductive inference of a large class of context-free languages using the paradigm of positive data and a membership oracle. We achieve this result by moving to a novel representation, called Contextual Binary Feature Grammars (CBFGs), which are capable of representing richly structured context-free languages as well as some context sensitive...
متن کاملLearning Nonregular Languages: A Comparison of Simple Recurrent Networks and LSTM
In response to Rodriguez's recent article (2001), we compare the performance of simple recurrent nets and long short-term memory recurrent nets on context-free and context-sensitive languages.
متن کاملEfficient, Correct, Unsupervised Learning for Context-Sensitive Languages
A central problem for NLP is grammar induction: the development of unsupervised learning algorithms for syntax. In this paper we present a lattice-theoretic representation for natural language syntax, called Distributional Lattice Grammars. These representations are objective or empiricist, based on a generalisation of distributional learning, and are capable of representing all regular languag...
متن کاملLSTM recurrent networks learn simple context-free and context-sensitive languages
Previous work on learning regular languages from exemplary training sequences showed that long short-term memory (LSTM) outperforms traditional recurrent neural networks (RNNs). We demonstrate LSTMs superior performance on context-free language benchmarks for RNNs, and show that it works even better than previous hardwired or highly specialized architectures. To the best of our knowledge, LSTM ...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 1999